The search functionality is under construction.

Keyword Search Result

[Keyword] cellular neural network(27hit)

21-27hit(27hit)

  • Discrete Time Cellular Neural Networks with Two Types of Neuron Circuits for Image Coding and Their VLSI Implementations

    Cong-Kha PHAM  Munemitsu IKEGAMI  Mamoru TANAKA  

     
    PAPER

      Vol:
    E78-A No:8
      Page(s):
    978-988

    This paper described discrete time Cellular Neural Networks (DT-CNN) with two types of neuron circuits for image coding from an analog format to a digital format and their VLSI implementations. The image coding methods proposed in this paper have been investigated for a purpose of transmission of a coded image and restoration again without a large loss of an original image information. Each neuron circuti of a network receives one pixel of an input image, and processes it with binary outputs data fed from neighboring neuron circuits. Parallel dynamics quantization methods have been adopted for image coding methods. They are performed in networks to decide an output binary value of each neuron circuit according to output values of neighboring neuron circuits. Delayed binary outputs of neuron circuits in a neighborhood are directly connected to inputs of a current active neuron circuit. Next state of a network is computed form a current state at some neuron circuits in any time interval. Models of two types of neuron circuits and networks are presented and simulated to confirm an ability of proposed methods. Also, physical layout designs of coding chips have been done to show their possibility of VLSI realizations.

  • Parallel Analog Image Coding and Decoding by Using Cellular Neural Networks

    Mamoru TANAKA  Kenneth R. CROUNSE  Tamás ROSKA  

     
    PAPER-Neural Networks

      Vol:
    E77-A No:8
      Page(s):
    1387-1395

    This paper describes highly parallel analog image coding and decoding by cellular neural networks (CNNs). The communication system in which the coder (C-) and decoder (D-) CNNs are embedded consists of a differential transmitter with an internal receiver model in the feedback loop. The C-CNN encodes the image through two cascaded techniques: structural compression and halftoning. The D-CNN decodes the received data through a reconstruction process, which includes a dynamic current distribution, so that the original input to the C-CNN can be recognized. The halftoning serves as a dynamic quantization to convert each pixel to a binary value depending on the neighboring values. We approach halftoning by the minimization of error energy between the original gray image and reconstructed halftone image, and the structural compression from the viewpoints of topological and regularization theories. All dynamics are described by CNN state equations. Both the proposed coding and decoding algorithms use only local image information in a space inveriant manner, therefore errors are distributed evenly and will not introduce the blocking effects found in DCT-based coding methods. In the future, the use of parallel inputs from on-chip photodetectors would allow direct dynamic quantization and compression of image sequences without the use of multiple bit analog-to-digital converters. To validate our theory, a simulation has been performed by using the relaxation method on an 150 frame image sequence. Each input image was 256256 pixels whth 8 bits per pixel. The simulated fixed compression rate, not including the Huffman coding, was about 1/16 with a PSNR of 31[dB]35[dB].

  • Pipelining Gauss Seidel Method for Analysis of Discrete Time Cellular Neural Networks

    Naohiko SHIMIZU  Gui-Xin CHENG  Munemitsu IKEGAMI  Yoshinori NAKAMURA  Mamoru TANAKA  

     
    PAPER-Neural Networks

      Vol:
    E77-A No:8
      Page(s):
    1396-1403

    This paper describes a pipelining universal system of discrete time cellular neural networks (DTCNNs). The new relaxation-based algorithm which is called a Pipelining Gauss Seidel (PGS) method is used to solve the CNN state equations in pipelining. In the systolic system of N processor elements {PEi}, each PEi performs the convolusional computation (CC) of all cells and the preceding PEi-1 performs the CC of all cells taking precedence over it by the precedence interval number p. The expected maximum number of PE's for the speeding up is given by n/p where n means the number of cells. For its application, the encoding and decoding process of moving images is simulated.

  • A Memory-Based Recurrent Neural Architecture for Chip Emulating Cortical Visual Processing

    Luigi RAFFO  Silvio P. SABATINI  Giacomo INDIVERI  Giovanni NATERI  Giacomo M. BISIO  

     
    PAPER

      Vol:
    E77-C No:7
      Page(s):
    1065-1074

    The paper describes the architecture and the simulated performances of a memory-based chip that emulates human cortical processing in early visual tasks, such as texture segregation. The featural elements present in an image are extracted by a convolution block and subsequently processed by the cortical chip, whose neurons, organized into three layers, gain relational descriptions (intelligent processing) through recurrent inhibitory/excitatory interactions between both inter-and intra-layer parallel pathways. The digital implementation of this architecuture directly maps the set of equations determining the status of the cortical network to achieve an optimal exploitation of VLSI technology in neural computation. Neurons are mapped into a memory matrix whose elements are updated through a programmable computational unit that implements synaptic interconnections. By using 0.5 µm-CMOS technology, full cortical image processing can be attained on a single chip (2020 mm2 die) at a rate higher than 70 frames/second, for images of 256256 pixels.

  • On a Unified Synthesizing Approach for Cellular Neural Networks

    Chun-ying HO  Shinsaku MORI  

     
    PAPER-Network Synthesis

      Vol:
    E77-D No:4
      Page(s):
    433-442

    In this paper, we develop a unified synthesizing approach for the cloning templates of Cellular Neural Networks (CNNs). In particular, we shall consider the case when the signal processing problem is complex, and a multilayered CNN with time-variant templates is necessary. The method originates from the existence of correspondence between the cloning templates of Cellular Neural Network and its discrete counterpart, Discrete-Time Cellular Neural Network (DTCNN), in solving a prescribed image processing problem when time-variant templates are involved. Thus, one can start with calculating the cloning templates from DTCNN, and then translating the cloning templates to those for CNN operations. As a result, the mathematical tools being used in the synthesis of Discrete-time Cellular Neural Network can also be applied to the analog type Cellular Neural Network. This inevitably helps to simplify the design problem of CNN for signal processing. Examples akin to contour drawing and parallel thinning are shown to illustrate the merits of our proposed method.

  • Iterative Middle Mapping Learning Algorithm for Cellular Neural Networks

    Chen HE  Akio USHIDA  

     
    PAPER-Neural Networks

      Vol:
    E77-A No:4
      Page(s):
    706-715

    In this paper, a middle-mapping learning algorithm for cellular associative memories is presented. This algorithm makes full use of the properties of the cellular neural network so that the associative memory has some advantages compared with the memory designed by the ourter product method. It can guarantee each prototype is stored at an equilibrium point. In the practical implementation, it is easy to build up the circuit because the weight matrix presenting the connection between cells is not symmetric. The synchronous updating rule makes its associative speed very fast compared to the Hopfield associative memory.

  • Image Compression and Regeneration by Nonlinear Associative Silicon Retina

    Mamoru TANAKA  Yoshinori NAKAMURA  Munemitsu IKEGAMI  Kikufumi KANDA  Taizou HATTORI  Yasutami CHIGUSA  Hikaru MIZUTANI  

     
    PAPER-Neural Systems

      Vol:
    E75-A No:5
      Page(s):
    586-594

    Threre are two types of nonlinear associative silicon retinas. One is a sparse Hopfield type neural network which is called a H-type retina and the other is its dual network which is called a DH-type retina. The input information sequences of H-type and HD-type retinas are given by nodes and links as voltages and currents respectively. The error correcting capacity (minimum basin of attraction) of H-type and DH-type retinas is decided by the minimum numbers of links of cutset and loop respectively. The operation principle of the regeneration is based on the voltage or current distribution of the neural field. The most important nonlinear operation in the retinas is a dynamic quantization to decide the binary value of each neuron output from the neighbor value. Also, the edge is emphasized by a line-process. The rates of compression of H-type and DH-type retinas used in the simulation are 1/8 and (2/3) (1/8) respectively, where 2/3 and 1/8 mean rates of the structural and binarizational compression respectively. We could have interesting and significant simulation results enough to make a chip.

21-27hit(27hit)